Goto

Collaborating Authors

 chaotic dynamic


Chaotic Dynamics are Intrinsic to Neural Network Training with SGD

Neural Information Processing Systems

With the advent of deep learning over the last decade, a considerable amount of effort has gone into better understanding and enhancing Stochastic Gradient Descent so as to improve the performance and stability of artificial neural network training. Active research fields in this area include exploiting second order information of the loss landscape and improving the understanding of chaotic dynamics in optimization. This paper exploits the theoretical connection between the curvature of the loss landscape and chaotic dynamics in neural network training to propose a modified SGD ensuring non-chaotic training dynamics to study the importance thereof in NN training. Building on this, we present empirical evidence suggesting that the negative eigenspectrum - and thus directions of local chaos - cannot be removed from SGD without hurting training performance. Extending our empirical analysis to long-term chaos dynamics, we challenge the widespread understanding of convergence against a confined region in parameter space. Our results show that although chaotic network behavior is mostly confined to the initial training phase, models perturbed upon initialization do diverge at a slow pace even after reaching top training performance, and that their divergence can be modelled through a composition of a random walk and a linear divergence. The tools and insights developed as part of our work contribute to improving the understanding of neural network training dynamics and provide a basis for future improvements of optimization methods.


Field-programmable dynamics in a soft magnetic actuator enabling true random number generation and reservoir computing

Oliveros-Mata, Eduardo Sergio, Pylypovskyi, Oleksandr V., Raimondo, Eleonora, Illing, Rico, Zabila, Yevhen, Guo, Lin, Mu, Guannan, López, Mónica Navarro, Wang, Xu, Tzortzinis, Georgios, Filippatos, Angelos, Bermúdez, Gilbert Santiago Cañón, Garescì, Francesca, Finocchio, Giovanni, Makarov, Denys

arXiv.org Artificial Intelligence

Department of Mathematical and Computer Sciences, Physical Sciences and Earth Sciences, University of Messina, 8166 Messina, Italy Complex and even chaotic dynamics, though prevalent in many natural and engineered systems, has been largely avoided in the design of electromechanical systems due to concerns about wear and controlability. Here, we demonstrate that complex dynamics might be particularly advantageous in soft robotics, offering new functionalities beyond motion not easily achievable with traditional actuation methods. We designed and realized resilient magnetic soft actuators capable of operating in a tunable dynamic regime for tens of thousands cycles without fatigue. We experimentally demonstrated the application of these actuators for true random number generation and stochastic computing. These findings show that exploring the complex dynamics in soft robotics would extend the application scenarios in soft computing, human-robot interaction and collaborative robots as we demonstrate with biomimetic blinking and randomized voice modulation. A large number of mechanical systems, including simple ones such as the double pendulum, exhibit dynamics characterized by deterministic periodic and chaotic responses depending on the excitation frequency f and amplitude A of the applied force [1]. Mechanical systems with a tendency to chaotisation demonstrate multiple resonances and various transitions to chaos [2]. Today, the concept of complexity and, especially, deterministic chaos that refers to systems without stochastic fluctuations jet losing stability of phase space trajectories is explored for a variety of directions [3] even including biological systems [4] or optics [5]. In particular, chaos is a fundamental aspect of electromechanical systems and is broadly explored in motion planning for mobile rigid robots, fluid mixing, and improving energy harvesting, as well as in mechanisms used in washing machines, dishwashers, and air conditioners [6]. Although the analysis of traditional robotics and mechanisms has revealed inherent chaotic dynamics [7], chaos can also be intentionally generated through nonlinear feedback [6] to achieve specific functionalities. In contrast to rigid mechanisms, soft actuators can facilitate transition into complex dynamics without the need for dedicated feedback algorithms. Mechanically soft actuators do not possess any rigid components in their embodiment rendering them ideally suited to explore complex and even chaotic dynamics which is typically observed at higher frequencies (Supplementary Tables 1 and 2). The inherent nonlinear oscillations emerging in soft actuators for specific parameter values [8, 9] can be applied for secure, biomimetic, and soft computing applications.


Chaotic Dynamics are Intrinsic to Neural Network Training with SGD

Neural Information Processing Systems

With the advent of deep learning over the last decade, a considerable amount of effort has gone into better understanding and enhancing Stochastic Gradient Descent so as to improve the performance and stability of artificial neural network training.



On the Generalisation of Koopman Representations for Chaotic System Control

Hjikakou, Kyriakos, Cartagena, Juan Diego Cardenas, Sabatelli, Matthia

arXiv.org Artificial Intelligence

This paper investigates the generalisability of Koopman-based representations for chaotic dynamical systems, focusing on their transferability across prediction and control tasks. Using the Lorenz system as a testbed, we propose a three-stage methodology: learning Koopman embeddings through autoencoding, pre-training a transformer on next-state prediction, and fine-tuning for safety-critical control. Our results show that Koopman embeddings outperform both standard and physics-informed PCA baselines, achieving accurate and data-efficient performance. Notably, fixing the pre-trained transformer weights during fine-tuning leads to no performance degradation, indicating that the learned representations capture reusable dynamical structure rather than task-specific patterns. These findings support the use of Koopman embeddings as a foundation for multi-task learning in physics-informed machine learning. A project page is available at https://kikisprdx.github.io/.


Brain-inspired Chaotic Graph Backpropagation for Large-scale Combinatorial Optimization

Tao, Peng, Aihara, Kazuyuki, Chen, Luonan

arXiv.org Artificial Intelligence

Graph neural networks (GNNs) with unsupervised learning can solve large-scale combinatorial optimization problems (COPs) with efficient time complexity, making them versatile for various applications. However, since this method maps the combinatorial optimization problem to the training process of a graph neural network, and the current mainstream backpropagation-based training algorithms are prone to fall into local minima, the optimization performance is still inferior to the current state-of-the-art (SOTA) COP methods. To address this issue, inspired by possibly chaotic dynamics of real brain learning, we introduce a chaotic training algorithm, i.e. chaotic graph backpropagation (CGBP), which introduces a local loss function in GNN that makes the training process not only chaotic but also highly efficient. Different from existing methods, we show that the global ergodicity and pseudo-randomness of such chaotic dynamics enable CGBP to learn each optimal GNN effectively and globally, thus solving the COP efficiently. We have applied CGBP to solve various COPs, such as the maximum independent set, maximum cut, and graph coloring. Results on several large-scale benchmark datasets showcase that CGBP can outperform not only existing GNN algorithms but also SOTA methods. In addition to solving large-scale COPs, CGBP as a universal learning algorithm for GNNs, i.e. as a plug-in unit, can be easily integrated into any existing method for improving the performance.


On the difficulty of learning chaotic dynamics with RNNs

Neural Information Processing Systems

Recurrent neural networks (RNNs) are wide-spread machine learning tools for modeling sequential and time series data. They are notoriously hard to train because their loss gradients backpropagated in time tend to saturate or diverge during training. This is known as the exploding and vanishing gradient problem. Previous solutions to this issue either built on rather complicated, purpose-engineered architectures with gated memory buffers, or - more recently - imposed constraints that ensure convergence to a fixed point or restrict (the eigenspectrum of) the recurrence matrix. Such constraints, however, convey severe limitations on the expressivity of the RNN.


Chaotic Dynamics are Intrinsic to Neural Network Training with SGD

Neural Information Processing Systems

With the advent of deep learning over the last decade, a considerable amount of effort has gone into better understanding and enhancing Stochastic Gradient Descent so as to improve the performance and stability of artificial neural network training. Active research fields in this area include exploiting second order information of the loss landscape and improving the understanding of chaotic dynamics in optimization. This paper exploits the theoretical connection between the curvature of the loss landscape and chaotic dynamics in neural network training to propose a modified SGD ensuring non-chaotic training dynamics to study the importance thereof in NN training. Building on this, we present empirical evidence suggesting that the negative eigenspectrum - and thus directions of local chaos - cannot be removed from SGD without hurting training performance. Extending our empirical analysis to long-term chaos dynamics, we challenge the widespread understanding of convergence against a confined region in parameter space.


Accurate deep learning-based filtering for chaotic dynamics by identifying instabilities without an ensemble

Bocquet, Marc, Farchi, Alban, Finn, Tobias S., Durand, Charlotte, Cheng, Sibo, Chen, Yumeng, Pasmans, Ivo, Carrassi, Alberto

arXiv.org Machine Learning

We investigate the ability to discover data assimilation (DA) schemes meant for chaotic dynamics with deep learning. The focus is on learning the analysis step of sequential DA, from state trajectories and their observations, using a simple residual convolutional neural network, while assuming the dynamics to be known. Experiments are performed with the Lorenz 96 dynamics, which display spatiotemporal chaos and for which solid benchmarks for DA performance exist. The accuracy of the states obtained from the learned analysis approaches that of the best possibly tuned ensemble Kalman filter, and is far better than that of variational DA alternatives. Critically, this can be achieved while propagating even just a single state in the forecast step. We investigate the reason for achieving ensemble filtering accuracy without an ensemble. We diagnose that the analysis scheme actually identifies key dynamical perturbations, mildly aligned with the unstable subspace, from the forecast state alone, without any ensemble-based covariances representation. This reveals that the analysis scheme has learned some multiplicative ergodic theorem associated to the DA process seen as a non-autonomous random dynamical system.


Cyclic image generation using chaotic dynamics

Tanaka, Takaya, Yamaguti, Yutaka

arXiv.org Artificial Intelligence

Successive image generation using cyclic transformations is demonstrated by extending the CycleGAN model to transform images among three different categories. Repeated application of the trained generators produces sequences of images that transition among the different categories. The generated image sequences occupy a more limited region of the image space compared with the original training dataset. Quantitative evaluation using precision and recall metrics indicates that the generated images have high quality but reduced diversity relative to the training dataset. Such successive generation processes are characterized as chaotic dynamics in terms of dynamical system theory. Positive Lyapunov exponents estimated from the generated trajectories confirm the presence of chaotic dynamics, with the Lyapunov dimension of the attractor found to be comparable to the intrinsic dimension of the training data manifold. The results suggest that chaotic dynamics in the image space defined by the deep generative model contribute to the diversity of the generated images, constituting a novel approach for multi-class image generation. This model can be interpreted as an extension of classical associative memory to perform hetero-association among image categories.